Information Divergence is more chi squared distributed than the chi squared statistics
نویسندگان
چکیده
For testing goodness of fit it is very popular to use either the χ-statistic or G-statistics (information divergence). Asymptotically both are χ-distributed so an obvious question is which of the two statistics that has a distribution that is closest to the χ-distribution. Surprisingly, when there is only one degree of freedom it seems like the distribution of information divergence is much better approximated by a χ-distribution than the χ-statistic. For random variables we introduce a new transformation that transform several important distributions into new random variables that are almost Gaussian. For the binomial distributions and the Poisson distributions we formulate a general conjecture about how close their transform are to the Gaussian. The conjecture is proved for Poisson distributions. I. CHOICE OF STATISTIC We consider the problem of testing goodness-of-fit in a discrete setting. Here we shall follow the classic approach to this problem as developed by Pearson, Neyman and Fisher. The question is whether a sample with observation counts (X1, X2, . . . , Xk) has been generated by the distribution Q = (q1, q2, . . . , qk) . For sample size n the counts (X1, X2, . . . , Xk) is assumed to have a multinomial distribution. We introduce the empirical distribution P̂ = ( X1 n , X2 n , . . . , Xk n ) where n denotes the sample size n = X1 +X2 + · · ·+Xk. Often one uses one of the Csiszár [1] f -divergences
منابع مشابه
-Divergences and Related Distances
Derivation of tight bounds on f -divergences and related distances is of interest in information theory and statistics. This paper improves some existing bounds on f -divergences. In some cases, an alternative approach leads to a simplified proof of an existing bound. Following bounds on the chi-squared divergence, an improved version of a reversed Pinsker’s inequality is derived for an arbitra...
متن کاملIRWIN AND JOAN JACOBS CENTER FOR COMMUNICATION AND INFORMATION TECHNOLOGIES On Improved Bounds for Probability Metrics and f- Divergences
Derivation of tight bounds for probability metrics and f -divergences is of interest in information theory and statistics. This paper provides elementary proofs that lead, in some cases, to significant improvements over existing bounds; they also lead to the derivation of some existing bounds in a simplified way. The inequalities derived in this paper relate between the Bhattacharyya parameter,...
متن کاملOn Improved Bounds for Probability Metrics and $f$-Divergences
Derivation of tight bounds for probability metrics and f -divergences is of interest in information theory and statistics. This paper provides elementary proofs that lead, in some cases, to significant improvements over existing bounds; they also lead to the derivation of some existing bounds in a simplified way. The inequalities derived in this paper relate between the Bhattacharyya parameter,...
متن کاملIRWIN AND JOAN JACOBS CENTER FOR COMMUNICATION AND INFORMATION TECHNOLOGIES Bounds on f-Divergences and Related Distances
Derivation of tight bounds on f -divergences and related distances is of interest in information theory and statistics. This paper improves some existing bounds on f -divergences. In some cases, an alternative approach leads to a simplified proof of an existing bound. Following bounds on the chi-squared divergence, an improved version of a reversed Pinsker’s inequality is derived for an arbitra...
متن کاملGoodness-of-fit Tests and Minimum Power Divergence Estimators for Survival Data
Power-divergence statistics are proposed for grouped survival data. They are analogous to the power-divergence family of statistics proposed and studied in detail by Read and Cressie (1988) and Cressie and Read (1984) for contingency tables. The proposed statistics are useful for testing validity of parametric model assumptions in analyses of survival data. It is shown that these statistics hav...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- CoRR
دوره abs/1202.1125 شماره
صفحات -
تاریخ انتشار 2012